Goto

Collaborating Authors

 algorithmic stability and generalization



Algorithmic stability and generalization of an unsupervised feature selection algorithm

Neural Information Processing Systems

Feature selection, as a vital dimension reduction technique, reduces data dimension by identifying an essential subset of input features, which can facilitate interpretable insights into learning and inference processes. Algorithmic stability is a key characteristic of an algorithm regarding its sensitivity to perturbations of input samples. In this paper, we propose an innovative unsupervised feature selection algorithm attaining this stability with provable guarantees. The architecture of our algorithm consists of a feature scorer and a feature selector. The scorer trains a neural network (NN) to globally score all the features, and the selector adopts a dependent sub-NN to locally evaluate the representation abilities for selecting features. Further, we present algorithmic stability analysis and show that our algorithm has a performance guarantee via a generalization error bound. Extensive experimental results on real-world datasets demonstrate superior generalization performance of our proposed algorithm to strong baseline methods. Also, the properties revealed by our theoretical analysis and the stability of our algorithm-selected features are empirically confirmed.


Supplementary Material of " Algorithmic Stability and Generalization of an Unsupervised Feature Selection Algorithm "

Neural Information Processing Systems

Correspondence should be addressed to: qiang.cheng@uky.edu. The architecture of our algorithm is shown in Figure 1. For the training based on Eq. (2) of the main text, in each iteration of backpropagation, After training, only the trained selector is used to select features and do reconstruction during testing time. In Eq. (2) of the main text, the second term helps obtain During testing time, only the trained sub-NN is used to select features and do reconstruction. It has 5, 744 samples and 561 features.


Algorithmic stability and generalization of an unsupervised feature selection algorithm

Neural Information Processing Systems

Feature selection, as a vital dimension reduction technique, reduces data dimension by identifying an essential subset of input features, which can facilitate interpretable insights into learning and inference processes. Algorithmic stability is a key characteristic of an algorithm regarding its sensitivity to perturbations of input samples. In this paper, we propose an innovative unsupervised feature selection algorithm attaining this stability with provable guarantees. The architecture of our algorithm consists of a feature scorer and a feature selector. The scorer trains a neural network (NN) to globally score all the features, and the selector adopts a dependent sub-NN to locally evaluate the representation abilities for selecting features.